Critical Points Of An Autoencoder Can Provably Recover Sparsely Used Overcomplete Dictionaries

نویسندگان

  • Akshay Rangamani
  • Anirbit Mukherjee
  • Ashish Arora
  • Tejaswini Ganapathy
  • Amitabh Basu
  • Sang Peter Chin
  • Trac D. Tran
چکیده

In Dictionary Learning one is trying to recover incoherent matrices A∗ ∈ Rn×h (typically overcomplete and whose columns are assumed to be normalized) and sparse vectors x∗ ∈ R with a small support of size h for some 0 < p < 1 while being given access to observations y ∈ R where y = A∗x∗. In this work we undertake a rigorous analysis of the possibility that dictionary learning could be performed by gradient descent on Autoencoders, which are R → R neural network with a single ReLU activation layer of size h. Towards the above objective we propose a new autoencoder loss function which modifies the squared loss error term and also adds new regularization terms. We create a proxy for the expected gradient of this loss function which we motivate with high probability arguments, under natural distributional assumptions on the sparse code x∗. Under the same distributional assumptions on x∗, we show that, in the limit of large enough sparse code dimension, any zero point of our proxy for the expected gradient of the loss function within a certain radius of A∗ corresponds to dictionaries whose action on the sparse vectors is indistinguishable from that of A∗. We also report simulations on synthetic data in support of our theory. ∗Equal Contribution †[email protected][email protected] [email protected][email protected][email protected] ∗∗[email protected] ††[email protected] 1 ar X iv :1 70 8. 03 73 5v 1 [ cs .L G ] 1 2 A ug 2 01 7

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Exact Recovery of Sparsely Used Overcomplete Dictionaries

We consider the problem of learning overcomplete dictionaries in the context of sparse coding, where each sample selects a sparse subset of dictionary elements. Our method consists of two stages, viz., initial estimation of the dictionary, and a clean-up phase involving estimation of the coefficient matrix, and re-estimation of the dictionary. We prove that our method exactly recovers both the ...

متن کامل

Learning Sparsely Used Overcomplete Dictionaries

We consider the problem of learning sparsely used overcomplete dictionaries, where each observation is a sparse combination of elements from an unknown overcomplete dictionary. We establish exact recovery when the dictionary elements are mutually incoherent. Our method consists of a clustering-based initialization step, which provides an approximate estimate of the true dictionary with guarante...

متن کامل

Dictionary Learning and Anti-Concentration Broadening the Reach of Efficient, Gradient-Descent Algorithms for Learning Sparsely-Used, Incoherent Dictionaries

As central as concentration of measure is to statistics and machine learning, this thesis aims to motivate anti-concentration as a promising and under-utilized toolkit for the design and analysis of statistical learning algorithms. This thesis focuses on learning incoherent dictionaries A∗ from observations y = A∗x, where x is a sparse coefficient vector drawn from a generative model. We impose...

متن کامل

Provably Accurate Double-Sparse Coding

Sparse coding is a crucial subroutine in algorithms for various signal processing, deep learning, and other machine learning applications. The central goal is to learn an overcomplete dictionary that can sparsely represent a given dataset. However, storage, transmission, and processing of the learned dictionary can be untenably high if the data dimension is high. In this paper, we consider the ...

متن کامل

Overcomplete Dictionary Design by Empirical Risk Minimization

Recently, there have been a growing interest in application of sparse representation for inverse problems. Most studies concentrated in devising ways for sparsely representing a solution using a given prototype overcomplete dictionary. Very few studies have addressed the more challenging problem of construction of an optimal overcomplete dictionary, and even these were primarily devoted to the ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • CoRR

دوره abs/1708.03735  شماره 

صفحات  -

تاریخ انتشار 2017